Today: Law of Large Numbers
Goal: start to understand error as it relates to sample size
Objectives:
- Distribution of the mean \(\bar{X}_{n}\)
- Chebyshev’s Inequality
- Law of Large Numbers
i.i.d.
Let us now seek the distribution of the mean \[\bar{X}_{n} = \displaystyle\frac{X_{1} + X_{2} + ... + X_{n}}{n}\]
- expected value
- variance
Distribution of Mean
Far from the Mean
- \(\mu\): population average
- \(a\): tolerance
Claim: \(P(|X - \mu| \geq a) \leq \displaystyle\frac{\text{Var}(X)}{a^{2}}\)
Chebyshev’s Inequality
That is, if we know the variance of a distribution, we can compute an upper bound for the probability of rare events!
Law of Large Numbers
The Law of Large Numbers basically combines Chebyshev’s Inequality with the earlier work for the distribution of the mean:
- \(\text{E}[\bar{X}_{n}] = \mu \quad\text{and}\quad \text{Var}(\bar{X}_{n}) = \displaystyle\frac{\sigma^{2}}{n}\)
- \(P(|X - \mu| \geq a) \leq \displaystyle\frac{\text{Var}(X)}{a^{2}}\)
Idea: What happens when we observe a lot of data?
That is, the probability that the mean of a sample of random variables is ``far’’ from the inherent expected value eventually goes to zero.
Nerdy Example
How many numbers between zero and one do we have to add up to have a sum that is greater than one?
- Let \(X_{i} \sim U(0,1)\) be i.i.d.
- Let \(Y\) be the amount of \(X_{i}\) added together to get a sum greater than one
- For a conservative estimate, suppose \(Y \sim U(2,6)\), then
\[\text{Var}(Y) = \displaystyle\frac{4}{3}\] - Empirically (i.e. computer simulation), we saw convergence toward
\[\bar{Y}_{n} = e \approx 2.7183\]
How many trials are needed so that the simulations converge to a mean within 0.01 of the true answer with at least 95 percent probability?
Looking Ahead
due Fri., Mar. 24:
- LHW8
no lecture on Mar. 24, Apr. 3
Exam 2 will be on Mon., Apr. 10